Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Quantization Update #150

Open
wants to merge 35 commits into
base: dev
Choose a base branch
from
Open

Quantization Update #150

wants to merge 35 commits into from

Conversation

sydarb
Copy link

@sydarb sydarb commented Sep 29, 2023

Proposed features:

  • refactor quantization codebase
  • add true quantization (8bit) support for all methods
  • implement AMP quantization (greedily for now) based on compression requirement

refer trailmet/algorithms/quantize/info.md for implementation details and remarks.
refer experiments/quantization/LAPQ/lapq_demo.ipynb for demo nb.

Do not merge until:

  • update requirements to pytorch v2.0
  • refactor quantization codebase to more simplified and modular
  • add automatic mixed precision quantization feature
  • add LAPQ true-quantization functionality
  • test LAPQ true-quantization and add demo notebook
  • test BRECQ fake-quantization with refactored changes
  • add BRECQ true-quantization functionality
  • test BRECQ true-quantization and add demo notebook
  • update and test quantized inverted residual module for mobilenet
  • fix logging
  • check naming and linting
  • update docstring and add comments

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

1 participant